Finding Latent Causes in Causal Networks: an Efficient Approach Based on Markov Blankets
نویسندگان
چکیده
Andre Elisseeff2 ae l@ zurich.ibm .com 2 Data Analytics Group IBM Research GmbH 8803 Rlischlikon, Switzerland Causal structure-discovery techniques usually assume that all causes of more than one variable are observed. This is the so-called causal sufficiency assumption. In practice, it is untestable, and often violated. In this paper, we present an efficient causal structure-learning algorithm, suited for causally insufficient data. Similar to algorithms such as IC* and FCI, the proposed approach drops the causal sufficiency assumption and learns a structure that indicates (potential) latent causes for pairs of observed variables. Assuming a constant local density of the data-generating graph, our algorithm makes a quadratic number of conditionalindependence tests w.r.t. the number of variables. We show with experiments that our algorithm is comparable to the state-of-the-art FCI algorithm in accuracy, while being several orders of magnitude faster on large problems. We conclude that MBCS* makes a new range of causally insufficient problems computationally tractable.
منابع مشابه
Bayesian Network Induction via Local Neighborhoods
In recent years, Bayesian networks have become highly successful tool for diagnosis, analysis, and decision making in real-world domains. We present an efficient algorithm for learning Bayes networks from data. Our approach constructs Bayesian networks by first identifying each node’s Markov blankets, then connecting nodes in a maximally consistent way. In contrast to the majority of work, whic...
متن کاملFast Gradient-Based Inference with Continuous Latent Variable Models in Auxiliary Form
We propose a technique for increasing the efficiency of gradient-based inference and learning in Bayesian networks with multiple layers of continuous latent variables. We show that, in many cases, it is possible to express such models in an auxiliary form, where continuous latent variables are conditionally deterministic given their parents and a set of independent auxiliary variables. Variable...
متن کاملUsing Markov Blankets for Causal Structure Learning
We show how a generic feature-selection algorithm returning strongly relevant variables can be turned into a causal structure-learning algorithm. We prove this under the Faithfulness assumption for the data distribution. In a causal graph, the strongly relevant variables for a node X are its parents, children, and children’s parents (or spouses), also known as the Markov blanket of X . Identify...
متن کاملLearning Markov Blankets for Continuous or Discrete Networks via Feature Selection
Markov Blankets discovery algorithms are important for learning a Bayesian network structure. We present an argument that tree ensemble masking measures can provide an approximate Markov blanket. Then an ensemble feature selection method is used to learn Markov blankets for either discrete or continuous networks (without linear, Gaussian assumptions). We compare our algorithm in the causal stru...
متن کاملExperimental Learning of Causal Models with Latent Variables
This article discusses graphical models that can handle latent variables without explicitly modeling them quantitatively. There exist several paradigms for such problem domains. Two of them are semi-Markovian causal models and maximal ancestral graphs. Applying these techniques to a problem domain consists of several steps, typically: structure learning from observational and experimental data,...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره شماره
صفحات -
تاریخ انتشار 2008